To Drop or Not to Drop: Robustness, Consistency and Differential Privacy Properties of Dropout

نویسندگان

  • Prateek Jain
  • Vivek Kulkarni
  • Abhradeep Thakurta
  • Oliver Williams
چکیده

Training deep belief networks (DBNs) requires optimizing a non-convex function with an extremely large number of parameters. Naturally, existing gradient descent (GD) based methods are prone to arbitrarily poor local minima. In this paper, we rigorously show that such local minima can be avoided (upto an approximation error) by using the dropout technique, a widely used heuristic in this domain. In particular, we show that by randomly dropping a few nodes of a one-hidden layer neural network, the training objective function, up to a certain approximation error, decreases by a multiplicative factor. On the flip side, we show that for training convex empirical risk minimizers (ERM), dropout in fact acts as a “stabilizer” or regularizer. That is, a simple dropout based GD method for convex ERMs is stable in the face of arbitrary changes to any one of the training points. Using the above assertion, we show that dropout provides fast rates for generalization error in learning (convex) generalized linear models (GLM). Moreover, using the above mentioned stability properties of dropout, we design dropout based differentially private algorithms for solving ERMs. The learned GLM thus, preserves privacy of each of the individual training points while providing accurate predictions for new test points. Finally, we empirically validate our stability assertions for dropout in the context of convex ERMs and show that surprisingly, dropout significantly outperforms (in terms of prediction accuracy) the L2 regularization based methods for several benchmark datasets. ar X iv :1 50 3. 02 03 1v 1 [ cs .L G ] 6 M ar 2 01 5

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Analysis on the Dropout Effect in Convolutional Neural Networks

Regularizing neural networks is an important task to reduce overfitting. Dropout [1] has been a widely-used regularization trick for neural networks. In convolutional neural networks (CNNs), dropout is usually applied to the fully connected layers. Meanwhile, the regularization effect of dropout in the convolutional layers has not been thoroughly analyzed in the literature. In this paper, we an...

متن کامل

Influence of Pipeline Operating Pressure on Value Setting of Automatic Control Valves at Different Pressure Drop Rates

When a natural gas pipeline ruptures, the adjacent upstream and downstream automatic control valves (ACV) should close quickly to prevent leakage or explosion. The differential pressure set point (DPS) at each valve location is the main criteria for value setting in ACV actions. If the DPS is not properly adjusted, the ACV may mistakenly close or it may not take any actions at a proper ti...

متن کامل

میزان عدم پیگیری درمان در درمانگاه سرپایی بیمارستان روزبه: پیگیری یکساله

Objectives: Mental illnesses are common among outpatient clinics, but only some of these patients receive adequate treatments and attend. The purpose of this study is to evaluate the dropout rate and identify the predictors of dropouts in outpatient clinics in Roozbeh Psychiatric Hospital. Methods: A questionnaire designed by expert psychiatrists was filled. The questionnaire consists of inf...

متن کامل

A Non-Random Dropout Model for Analyzing Longitudinal Skew-Normal Response

In this paper, multivariate skew-normal distribution is em- ployed for analyzing an outcome based dropout model for repeated mea- surements with non-random dropout in skew regression data sets. A probit regression is considered as the conditional probability of an ob- servation to be missing given outcomes. A simulation study of using the proposed methodology and comparing it with a semi-parame...

متن کامل

Experimental values for adjusting an automatic control valve in gas pipeline transportation

When a natural gas pipeline ruptures, the adjacent automatic line control valves (ALCVs) should close quickly to prevent leakage or explosion. The differential pressure set point (DPS) at each valve location is the main criteria for value setting in ALCV action. If the DPS is not properly adjusted, the ALCV may mistakenly close or it may not take any action at proper time. This study focused on...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • CoRR

دوره abs/1503.02031  شماره 

صفحات  -

تاریخ انتشار 2015